Northbrook
Multi-Agent Reinforcement Learning for Deadlock Handling among Autonomous Mobile Robots
This dissertation explores the application of multi-agent reinforcement learning (MARL) for handling deadlocks in intralogistics systems that rely on autonomous mobile robots (AMRs). AMRs enhance operational flexibility but also increase the risk of deadlocks, which degrade system throughput and reliability. Existing approaches often neglect deadlock handling in the planning phase and rely on rigid control rules that cannot adapt to dynamic operational conditions. To address these shortcomings, this work develops a structured methodology for integrating MARL into logistics planning and operational control. It introduces reference models that explicitly consider deadlock-capable multi-agent pathfinding (MAPF) problems, enabling systematic evaluation of MARL strategies. Using grid-based environments and an external simulation software, the study compares traditional deadlock handling strategies with MARL-based solutions, focusing on PPO and IMPALA algorithms under different training and execution modes. Findings reveal that MARL-based strategies, particularly when combined with centralized training and decentralized execution (CTDE), outperform rule-based methods in complex, congested environments. In simpler environments or those with ample spatial freedom, rule-based methods remain competitive due to their lower computational demands. These results highlight that MARL provides a flexible and scalable solution for deadlock handling in dynamic intralogistics scenarios, but requires careful tailoring to the operational context.
- North America > United States > New Jersey > Hudson County > Hoboken (0.13)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.13)
- North America > United States > New Jersey > Middlesex County > Piscataway (0.04)
- (29 more...)
- Research Report > New Finding (1.00)
- Overview (1.00)
- Workflow (0.92)
- Information Technology (1.00)
- Government (1.00)
- Education (1.00)
- (6 more...)
ROS-related Robotic Systems Development with V-model-based Application of MeROS Metamodel
Winiarski, Tomasz, Kaniuka, Jan, Giełdowski, Daniel, Ostrysz, Jakub, Radlak, Krystian, Kushnir, Dmytro
Systems built on the Robot Operating System (ROS) are increasingly easy to assemble, yet hard to govern and reliably coordinate. Beyond the sheer number of subsystems involved, the difficulty stems from their diversity and interaction depth. In this paper, we use a compact heterogeneous robotic system (HeROS), combining mobile and manipulation capabilities, as a demonstration vehicle under dynamically changing tasks. Notably, all its subsystems are powered by ROS. The use of compatible interfaces and other ROS integration capabilities simplifies the construction of such systems. However, this only addresses part of the complexity: the semantic coherence and structural traceability are even more important for precise coordination and call for deliberate engineering methods. The Model-Based Systems Engineering (MBSE) discipline, which emerged from the experience of complexity management in large-scale engineering domains, offers the methodological foundations needed. Despite their strengths in complementary aspects of robotics systems engineering, the lack of a unified approach to integrate ROS and MBSE hinders the full potential of these tools. Motivated by the anticipated impact of such a synergy in robotics practice, we propose a structured methodology based on MeROS - a SysML metamodel created specifically to put the ROS-based systems into the focus of the MBSE workflow. As its methodological backbone, we adapt the well-known V-model to this context, illustrating how complex robotic systems can be designed with traceability and validation capabilities embedded into their lifecycle using practices familiar to engineering teams.
- Europe > Poland > Masovia Province > Warsaw (0.04)
- Europe > Switzerland > Geneva > Geneva (0.04)
- Europe > Germany (0.04)
- (15 more...)
- Research Report (0.64)
- Workflow (0.49)
Auto-RAG: Autonomous Retrieval-Augmented Generation for Large Language Models
Yu, Tian, Zhang, Shaolei, Feng, Yang
Iterative retrieval refers to the process in which the model continuously queries the retriever during generation to enhance the relevance of the retrieved knowledge, thereby improving the performance of Retrieval-Augmented Generation (RAG). Existing work typically employs few-shot prompting or manually constructed rules to implement iterative retrieval. This introduces additional inference overhead and overlooks the remarkable reasoning capabilities of Large Language Models (LLMs). In this paper, we introduce Auto-RAG, an autonomous iterative retrieval model centered on the LLM's powerful decision-making capabilities. Auto-RAG engages in multi-turn dialogues with the retriever, systematically planning retrievals and refining queries to acquire valuable knowledge. This process continues until sufficient external information is gathered, at which point the results are presented to the user. To this end, we develop a method for autonomously synthesizing reasoning-based decision-making instructions in iterative retrieval and fine-tuned the latest open-source LLMs. The experimental results indicate that Auto-RAG is capable of autonomous iterative interaction with the retriever, effectively leveraging the remarkable reasoning and decision-making abilities of LLMs, which lead to outstanding performance across six benchmarks. Further analysis reveals that Auto-RAG can autonomously adjust the number of iterations based on the difficulty of the questions and the utility of the retrieved knowledge, without requiring any human intervention. Moreover, Auto-RAG expresses the iterative retrieval process in natural language, enhancing interpretability while providing users with a more intuitive experience\footnote{Code is available at \url{https://github.com/ictnlp/Auto-RAG}.
- Europe > Serbia (0.05)
- Asia > India > Maharashtra (0.04)
- North America > Mexico > Mexico City > Mexico City (0.04)
- (19 more...)
- Personal > Obituary (0.46)
- Research Report > New Finding (0.34)
- Media > Film (1.00)
- Leisure & Entertainment (1.00)
Video to Video Generative Adversarial Network for Few-shot Learning Based on Policy Gradient
Ma, Yintai, Klabjan, Diego, Utke, Jean
The development of sophisticated models for video-to-video synthesis has been facilitated by recent advances in deep reinforcement learning and generative adversarial networks (GANs). In this paper, we propose RL-V2V-GAN, a new deep neural network approach based on reinforcement learning for unsupervised conditional video-to-video synthesis. While preserving the unique style of the source video domain, our approach aims to learn a mapping from a source video domain to a target video domain. We train the model using policy gradient and employ ConvLSTM layers to capture the spatial and temporal information by designing a fine-grained GAN architecture and incorporating spatio-temporal adversarial goals. The adversarial losses aid in content translation while preserving style. Unlike traditional video-to-video synthesis methods requiring paired inputs, our proposed approach is more general because it does not require paired inputs. Thus, when dealing with limited videos in the target domain, i.e., few-shot learning, it is particularly effective. Our experiments show that RL-V2V-GAN can produce temporally coherent video results. These results highlight the potential of our approach for further advances in video-to-video synthesis.
- North America > United States > Illinois > Cook County > Northbrook (0.04)
- North America > United States > Illinois > Cook County > Evanston (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
The Open Autonomy Safety Case Framework
Wagner, Michael, Carlan, Carmen
A system safety case is a compelling, comprehensible, and valid argument about the satisfaction of the safety goals of a given system operating in a given environment supported by convincing evidence. Since the publication of UL 4600 in 2020, safety cases have become a best practice for measuring, managing, and communicating the safety of autonomous vehicles (AVs). Although UL 4600 provides guidance on how to build the safety case for an AV, the complexity of AVs and their operating environments, the novelty of the used technology, the need for complying with various regulations and technical standards, and for addressing cybersecurity concerns and ethical considerations make the development of safety cases for AVs challenging. To this end, safety case frameworks have been proposed that bring strategies, argument templates, and other guidance together to support the development of a safety case. This paper introduces the Open Autonomy Safety Case Framework, developed over years of work with the autonomous vehicle industry, as a roadmap for how AVs can be deployed safely and responsibly.
- North America > United States > New York (0.04)
- North America > United States > Texas > Travis County > Austin (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- (3 more...)
- Government > Regional Government > North America Government > United States Government (1.00)
- Automobiles & Trucks (1.00)
- Transportation > Ground > Road (0.93)
- (4 more...)
Physics of Language Models: Part 3.1, Knowledge Storage and Extraction
Allen-Zhu, Zeyuan, Li, Yuanzhi
Large language models (LLMs) can store a vast amount of world knowledge, often extractable via question-answering (e.g., "What is Abraham Lincoln's birthday?"). However, do they answer such questions based on exposure to similar questions during training (i.e., cheating), or by genuinely learning to extract knowledge from sources like Wikipedia? In this paper, we investigate this issue using a controlled biography dataset. We find a strong correlation between the model's ability to extract knowledge and various diversity measures of the training data. $\textbf{Essentially}$, for knowledge to be reliably extracted, it must be sufficiently augmented (e.g., through paraphrasing, sentence shuffling) $\textit{during pretraining}$. Without such augmentation, knowledge may be memorized but not extractable, leading to 0% accuracy, regardless of subsequent instruction fine-tuning. To understand why this occurs, we employ (nearly) linear probing to demonstrate a strong connection between the observed correlation and how the model internally encodes knowledge -- whether it is linearly encoded in the hidden embeddings of entity names or distributed across other token embeddings in the training text. This paper provides $\textbf{several key recommendations for LLM pretraining in the industry}$: (1) rewrite the pretraining data -- using small, auxiliary models -- to provide knowledge augmentation, and (2) incorporate more instruction-finetuning data into the pretraining stage before it becomes too late.
- North America > United States > New York > New York County > New York City (0.14)
- North America > United States > New York > Kings County > New York City (0.14)
- North America > United States > Wisconsin > Waukesha County > Menomonee Falls (0.14)
- (22 more...)
A Stochastic Approach to Classification Error Estimates in Convolutional Neural Networks
Peleska, Jan, Brüning, Felix, Gleirscher, Mario, Huang, Wen-ling
This technical report presents research results achieved in the field of verification of trained Convolutional Neural Network (CNN) used for image classification in safety-critical applications. As running example, we use the obstacle detection function needed in future autonomous freight trains with Grade of Automation (GoA) 4. It is shown that systems like GoA 4 freight trains are indeed certifiable today with new standards like ANSI/UL 4600 and ISO 21448 used in addition to the long-existing standards EN 50128 and EN 50129. Moreover, we present a quantitative analysis of the system-level hazard rate to be expected from an obstacle detection function. It is shown that using sensor/perceptor fusion, the fused detection system can meet the tolerable hazard rate deemed to be acceptable for the safety integrity level to be applied (SIL-3). A mathematical analysis of CNN models is performed which results in the identification of classification clusters and equivalence classes partitioning the image input space of the CNN. These clusters and classes are used to introduce a novel statistical testing method for determining the residual error probability of a trained CNN and an associated upper confidence limit. We argue that this greybox approach to CNN verification, taking into account the CNN model's internal structure, is essential for justifying that the statistical tests have covered the trained CNN with its neurons and inter-layer mappings in a comprehensive way.
- Europe > Germany > Bremen > Bremen (0.14)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (16 more...)
- Research Report > New Finding (1.00)
- Instructional Material (0.93)
- Research Report > Promising Solution (0.67)
- Transportation > Infrastructure & Services (1.00)
- Transportation > Ground > Rail (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
ICTSurF: Implicit Continuous-Time Survival Functions with Neural Networks
Puttanawarut, Chanon, Looareesuwan, Panu, Wabina, Romen Samuel, Saowaprut, Prut
Survival analysis, also known as time-to-event analysis, aims at estimating the survival distributions of a specific event and time-of-interests. Typically, the estimation of survival probability involves modeling a relationship between covariates and time-to-event that is typically partially observed; e.g., it may not be possible to observe the event status of the same sample. This presents one of the key challenges in the field of survival analysis. The conventional approaches commonly employed in survival analysis include the Cox Proportional Hazards (CPH) model, as proposed by Cox [6]. Although the CPH model is widely used, it is burdened by a substantial assumption of a consistent proportional hazard throughout the entire lifespan and a predetermined relationship between covariates. Other conventional methods, such as Weibull or Log-Normal distribution, also model a relationship between time and covariates based on a strong parametric assumption. Recently, due to the success of DNN-based models, the majority of research in survival analysis has shifted towards models built on DNNs, demonstrating superior performance compared to traditional approaches. Recent studies have shown that the majority of the survival models are an extension of the conventional CPH model [28].
- Asia > Thailand > Bangkok > Bangkok (0.05)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- North America > United States > Illinois > Cook County > Northbrook (0.04)
- Research Report > Experimental Study (0.48)
- Research Report > Strength Medium (0.34)
Automated Measurement of Vascular Calcification in Femoral Endarterectomy Patients Using Deep Learning
Rajeoni, Alireza Bagheri, Pederson, Breanna, Clair, Daniel G., Lessner, Susan M., Valafar, Homayoun
Atherosclerosis, a chronic inflammatory disease affecting the large arteries, presents a global health risk. Accurate analysis of diagnostic images, like computed tomographic angiograms (CTAs), is essential for staging and monitoring the progression of atherosclerosis-related conditions, including peripheral arterial disease (PAD). However, manual analysis of CTA images is time-consuming and tedious. To address this limitation, we employed a deep learning model to segment the vascular system in CTA images of PAD patients undergoing femoral endarterectomy surgery and to measure vascular calcification from the left renal artery to the patella. Utilizing proprietary CTA images of 27 patients undergoing femoral endarterectomy surgery provided by Prisma Health Midlands, we developed a Deep Neural Network (DNN) model to first segment the arterial system, starting from the descending aorta to the patella, and second, to provide a metric of arterial calcification. Our designed DNN achieved 83.4% average Dice accuracy in segmenting arteries from aorta to patella, advancing the state-of-the-art by 0.8%. Furthermore, our work is the first to present a robust statistical analysis of automated calcification measurement in the lower extremities using deep learning, attaining a Mean Absolute Percentage Error (MAPE) of 9.5% and a correlation coefficient of 0.978 between automated and manual calcification scores. These findings underscore the potential of deep learning techniques as a rapid and accurate tool for medical professionals to assess calcification in the abdominal aorta and its branches above the patella. The developed DNN model and related documentation in this project are available at GitHub page at https://github.com/pip-alireza/DeepCalcScoring.
- North America > United States > South Carolina > Richland County > Columbia (0.14)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Texas > Harris County > Houston (0.04)
- (8 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
BEND: Benchmarking DNA Language Models on biologically meaningful tasks
Marin, Frederikke Isa, Teufel, Felix, Horlacher, Marc, Madsen, Dennis, Pultz, Dennis, Winther, Ole, Boomsma, Wouter
The genome sequence contains the blueprint for governing cellular processes. While the availability of genomes has vastly increased over the last decades, experimental annotation of the various functional, non-coding and regulatory elements encoded in the DNA sequence remains both expensive and challenging. This has sparked interest in unsupervised language modeling of genomic DNA, a paradigm that has seen great success for protein sequence data. Although various DNA language models have been proposed, evaluation tasks often differ between individual works, and might not fully recapitulate the fundamental challenges of genome annotation, including the length, scale and sparsity of the data. In this study, we introduce BEND, a Benchmark for DNA language models, featuring a collection of realistic and biologically meaningful downstream tasks defined on the human genome. We find that embeddings from current DNA LMs can approach performance of expert methods on some tasks, but only capture limited information about long-range features. BEND is available at https://github.com/frederikkemarin/BEND.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.14)
- North America > United States > Texas > Harris County > Houston (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- (33 more...)